Goto

Collaborating Authors

 computational learning theory









Stability & Generalisation of Gradient Descent for Shallow Neural Networks without the Neural Tangent Kernel

Neural Information Processing Systems

This issue has attracted considerable attention in recent years with numerous works [Du et al., 2018, Lee et al., 2019, Allen-Zhu et al., 2019, Oymak and Soltanolkotabi, 2020] demonstrating that overparameterised shallow networks (in a sense


Erzeugunsgrad, VC-Dimension and Neural Networks with rational activation function

Pardo, Luis Miguel, Sebastián, Daniel

arXiv.org Artificial Intelligence

The notion of Erzeugungsgrad was introduced by Joos Heintz in 1983 to bound the number of non-empty cells occurring after a process of quantifier elimination. We extend this notion and the combinatorial bounds of Theorem 2 in Heintz (1983) using the degree for constructible sets defined in Pardo-Sebastián (2022). We show that the Erzeugungsgrad is the key ingredient to connect affine Intersection Theory over algebraically closed fields and the VC-Theory of Computational Learning Theory for families of classifiers given by parameterized families of constructible sets. In particular, we prove that the VC-dimension and the Krull dimension are linearly related up to logarithmic factors based on Intersection Theory. Using this relation, we study the density of correct test sequences in evasive varieties. We apply these ideas to analyze parameterized families of neural networks with rational activation function.


Collaborative PAC Learning

Avrim Blum, Nika Haghtalab, Ariel D. Procaccia, Mingda Qiao

Neural Information Processing Systems

We consider a collaborative PAC learning model, in which k players attempt to learn the same underlying concept. We ask how much more information is required to learn an accurate classifier for all players simultaneously. We refer to the ratio between the sample complexity of collaborative PAC learning and its non-collaborative (single-player) counterpart as the overhead.